AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Small-scale pre-training

# Small-scale pre-training

Arshstory
MIT
A 500-million-parameter text generation model based on the Llama architecture, specifically designed for story creation.
Text Generation Transformers
A
arshiaafshani
131
1
OPENCLIP SigLIP Tiny 14 Distill SigLIP 400m Cc9m
MIT
A lightweight vision-language model based on the SigLIP architecture, extracting knowledge from the larger SigLIP-400m model through distillation techniques, suitable for zero-shot image classification tasks.
Image Classification
O
PumeTu
30
0
Roberta Base 100M 3
RoBERTa variants pre-trained on datasets ranging from 1M to 1B tokens, including BASE and MED-SMALL specifications, suitable for natural language processing tasks in resource-limited scenarios
Large Language Model
R
nyu-mll
18
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase